Personalization in Federated Learning (FL) aims to modify a collaboratively trained global model according to each client. Current approaches to personalization in FL are at a coarse granularity, i.e. all the input instances of a client use the same personalized model. This ignores the fact that some instances are more accurately handled by the global model due to better generalizability. To address this challenge, this work proposes Flow, a fine-grained stateless personalized FL approach. Flow creates dynamic personalized models by learning a routing mechanism that determines whether an input instance prefers the local parameters or its global counterpart. Thus, Flow introduces per-instance routing in addition to leveraging per-client personalization to improve accuracies at each client. Further, Flow is stateless which makes it unnecessary for a client to retain its personalized state across FL rounds. This makes Flow practical for large-scale FL settings and friendly to newly joined clients. Evaluations on Stackoverflow, Reddit, and EMNIST datasets demonstrate the superiority in prediction accuracy of Flow over state-of-the-art non-personalized and only per-client personalized approaches to FL.
translated by 谷歌翻译
In vision and linguistics; the main input modalities are facial expressions, speech patterns, and the words uttered. The issue with analysis of any one mode of expression (Visual, Verbal or Vocal) is that lot of contextual information can get lost. This asks researchers to inspect multiple modalities to get a thorough understanding of the cross-modal dependencies and temporal context of the situation to analyze the expression. This work attempts at preserving the long-range dependencies within and across different modalities, which would be bottle-necked by the use of recurrent networks and adds the concept of delta-attention to focus on local differences per modality to capture the idiosyncrasy of different people. We explore a cross-attention fusion technique to get the global view of the emotion expressed through these delta-self-attended modalities, in order to fuse all the local nuances and global context together. The addition of attention is new to the multi-modal fusion field and currently being scrutinized for on what stage the attention mechanism should be used, this work achieves competitive accuracy for overall and per-class classification which is close to the current state-of-the-art with almost half number of parameters.
translated by 谷歌翻译
在混合神经网络中,昂贵的卷积层被不可训练的固定变换所取代,参数大幅减少。在以前的作品中,通过用小波代替卷积来获得良好的结果。然而,基于小波的混合网络继承了小波沿曲线及其轴偏置的消失力矩。我们建议使用剪力岩对重要图像功能(例如边缘,脊和斑点)的强大支持。最终的网络称为复杂的剪切网络(COSHNET)。它在针对Resnet-50和Resnet-18的时装摄影师上进行了测试,分别获得了92.2%和90.7%和91.8%的测试。所提出的网络具有49.9k参数,而RESNET-18的参数为11.18m,使用较少的拖鞋52倍。最后,我们在Resnet要求的200个时期与200个时期进行了培训,不需要任何高参数调整或正则化。代码:https://github.com/ujjawal-k-panchal/coshnet
translated by 谷歌翻译